# 4bit Quantization
Bge Multilingual Gemma2 GPTQ
Apache-2.0
This is the 4-bit GPTQ quantized version of the BAAI/bge-multilingual-gemma2 model, supporting multilingual text embedding tasks.
Text Embedding
Transformers

B
shuyuej
34
5
Traceback 12b
Apache-2.0
TraceBack 12b is a 4bit quantized version based on the Mistral-Nemo-Instruct architecture, focusing on instruction-following and chain-of-thought reasoning tasks.
Large Language Model
Transformers

T
secemp9
1,470
29
Universal NER UniNER 7B All Bnb 4bit Smashed
PrunaAI's compressed version of the UniNER-7B-all model, significantly reducing memory usage and energy consumption through quantization techniques while maintaining good named entity recognition capabilities.
Large Language Model
Transformers

U
PrunaAI
22
1
Featured Recommended AI Models